Computational Machine Learning in Theory and Praxis Produced as Part of the Esprit Working Group in Neural and Computational Learning, Neurocolt 8556

نویسنده

  • Ming Li
چکیده

In the last few decades a computational approach to machine learning has emerged based on paradigms from recursion theory and the theory of computation. Such ideas include learning in the limit, learning by enumer-ation, and probably approximately correct (pac) learning. These models usually are not suitable in practical situations. In contrast, statistics based inference methods have enjoyed a long and distinguished career. Currently, Bayesian reasoning in various forms, minimum message length (MML) and minimum description length (MDL), are widely applied approaches. They are the tools to use with particular machine learning praxis such as simulated annealing, genetic algorithms, genetic programming, artiicial neural networks, and the like. These statistical inference methods select the hypothesis which minimizes the sum of the length of the description of the hypothesis (also called`model') and the length of the description of the data relative to the hypothesis. It appears to us that the future of computational machine learning will include combinations of the approaches above coupled with guaranties with respect to used time and memory resources. Computational learning theory will move closer to practice and the application of the principles such as MDL require further justiication. Here, we survey some of the actors in this dichotomy between theory and praxis, we justify MDL via the Bayesian approach, and give a comparison between pac learning and MDL learning of decision trees.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Computing the Maximum Bichromatic Discrepancy, with Applications to Computer Graphics and Machine Learning Produced as Part of the Esprit Working Group in Neural and Computational Learning, Neurocolt 8556

Computing the maximum bichromatic discrepancy is an interesting theoretical problem with important applications in computational learning theory, computational geometry and computer graphics. In this paper we give algorithms to compute the maximum bichromatic discrepancy for simple geometric ranges, including rectangles and halfspaces. In addition, we give extensions to other discrepancy problems.

متن کامل

Decision Trees Have Approximate Fingerprints Produced as Part of the Esprit Working Group in Neural and Computational Learning, Neurocolt 8556

We prove that decision trees exhibit the \approximate ngerprint" property, and therefore are not polynomially learnable using only equivalence queries. A slight modiication of the proof extends this result to several other representation classes of boolean concepts which have been studied in computational learning theory.

متن کامل

Probabilistic Analysis of Learning in Artiicial Neural Networks: the Pac Model and Its Variants Produced as Part of the Esprit Working Group in Neural and Computational Learning, Neurocolt 8556

1 1 A version of this is to appear as a chapter in The Computational and Learning Complexity of Neural Networks (ed. Ian Parberry), MIT Press. 2 Abstract There are a number of mathematical approaches to the study of learning and generalization in artiicial neural networks. Here we survey thèprobably approximately correct' (PAC) model of learning and some of its variants. These models, much-stud...

متن کامل

Neural Networks with Quadratic Vc Dimension Produced as Part of the Esprit Working Group in Neural and Computational Learning, Neurocolt 8556 Submitted to Workshop on Neural Information Processing, Nips'95

This paper shows that neural networks which use continuous activation functions have VC dimension at least as large as the square of the number of weights w. This result settles a long-standing open question, namely whether the well-known O(w log w) bound, known for hard-threshold nets, also held for more general sigmoidal nets. Implications for the number of samples needed for valid generaliza...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1995